78 research outputs found

    Do optimization methods in deep learning applications matter?

    Get PDF
    With advances in deep learning, exponential data growth and increasing model complexity, developing efficient optimization methods are attracting much research attention. Several implementations favor the use of Conjugate Gradient (CG) and Stochastic Gradient Descent (SGD) as being practical and elegant solutions to achieve quick convergence, however, these optimization processes also present many limitations in learning across deep learning applications. Recent research is exploring higher-order optimization functions as better approaches, but these present very complex computational challenges for practical use. Comparing first and higher-order optimization functions, in this paper, our experiments reveal that Levemberg-Marquardt (LM) significantly supersedes optimal convergence but suffers from very large processing time increasing the training complexity of both, classification and reinforcement learning problems. Our experiments compare off-the-shelf optimization functions(CG, SGD, LM and L-BFGS) in standard CIFAR, MNIST, CartPole and FlappyBird experiments.The paper presents arguments on which optimization functions to use and further, which functions would benefit from parallelization efforts to improve pretraining time and learning rate convergence

    X-Machines for Agent-Based Modeling

    Get PDF
    This book discusses various aspects of agent-based modeling and simulation using FLAME (Flexible Large-scale Agent-Based Modeling Environment) which is a popular agent-based modeling environment that enables automatic parallelization of models. Along with a focus on the software engineering principles in building agent-based models, the book comprehensively discusses how models can be written for various domains including biology, economics and social networks. The book also includes examples to guide readers on how to write their own models
    corecore